1000 resultados para aggregation rules


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aggregation operator have been considered from a computable point of view. The important condition that the computation is friendly when portions of data are inserted o deleted to the list of values to aggregate is considered.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

In social choice theory, preference aggregation refers to computing an aggregate preference over a set of alternatives given individual preferences of all the agents. In real-world scenarios, it may not be feasible to gather preferences from all the agents. Moreover, determining the aggregate preference is computationally intensive. In this paper, we show that the aggregate preference of the agents in a social network can be computed efficiently and with sufficient accuracy using preferences elicited from a small subset of critical nodes in the network. Our methodology uses a model developed based on real-world data obtained using a survey on human subjects, and exploits network structure and homophily of relationships. Our approach guarantees good performance for aggregation rules that satisfy a property which we call expected weak insensitivity. We demonstrate empirically that many practically relevant aggregation rules satisfy this property. We also show that two natural objective functions in this context satisfy certain properties, which makes our methodology attractive for scalable preference aggregation over large scale social networks. We conclude that our approach is superior to random polling while aggregating preferences related to individualistic metrics, whereas random polling is acceptable in the case of social metrics.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The Water Framework Directive uses the “One-out, all-out” (OAOO) principle in assessing water bodies (i.e. the worst status of the elements used in the assessment determines the final status of the water body). Combination of multiple parameters within a biological quality element (BQEs) can be done in different ways. This study analysed several aggregation conditions within the BQE "Flora other than phytoplankton" (intertidal macroalgae, subtidal macroalgae, eelgrass beds and opportunistic blooms) using monitoring data collected along the Channel and Atlantic coastline. Four aggregation criteria were tested on two sets of data collected between 2004 and 2014: OOAO, average, intermediate method between OOAO and average and a method taking into account an uncertainty value at the threshold "Good/Moderate." Based on available data, the intermediate method appears the most qualified method using first an averaging approach between the natural habitat elements and then applying the OAOO between this mean and the opportunistic blooms, characteristic of an eutrophic environment. Expert judment might be used to ensure in the overall interpretation of results at waterbody level and in the classification outcomes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Many cell types form clumps or aggregates when cultured in vitro through a variety of mechanisms including rapid cell proliferation, chemotaxis, or direct cell-to-cell contact. In this paper we develop an agent-based model to explore the formation of aggregates in cultures where cells are initially distributed uniformly, at random, on a two-dimensional substrate. Our model includes unbiased random cell motion, together with two mechanisms which can produce cell aggregates: (i) rapid cell proliferation, and (ii) a biased cell motility mechanism where cells can sense other cells within a finite range, and will tend to move towards areas with higher numbers of cells. We then introduce a pair-correlation function which allows us to quantify aspects of the spatial patterns produced by our agent-based model. In particular, these pair-correlation functions are able to detect differences between domains populated uniformly at random (i.e. at the exclusion complete spatial randomness (ECSR) state) and those where the proliferation and biased motion rules have been employed - even when such differences are not obvious to the naked eye. The pair-correlation function can also detect the emergence of a characteristic inter-aggregate distance which occurs when the biased motion mechanism is dominant, and is not observed when cell proliferation is the main mechanism of aggregate formation. This suggests that applying the pair-correlation function to experimental images of cell aggregates may provide information about the mechanism associated with observed aggregates. As a proof of concept, we perform such analysis for images of cancer cell aggregates, which are known to be associated with rapid proliferation. The results of our analysis are consistent with the predictions of the proliferation-based simulations, which supports the potential usefulness of pair correlation functions for providing insight into the mechanisms of aggregate formation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The solution of a bivariate population balance equation (PBE) for aggregation of particles necessitates a large 2-d domain to be covered. A correspondingly large number of discretized equations for particle populations on pivots (representative sizes for bins) are solved, although at the end only a relatively small number of pivots are found to participate in the evolution process. In the present work, we initiate solution of the governing PBE on a small set of pivots that can represent the initial size distribution. New pivots are added to expand the computational domain in directions in which the evolving size distribution advances. A self-sufficient set of rules is developed to automate the addition of pivots, taken from an underlying X-grid formed by intersection of the lines of constant composition and constant particle mass. In order to test the robustness of the rule-set, simulations carried out with pivotwise expansion of X-grid are compared with those obtained using sufficiently large fixed X-grids for a number of composition independent and composition dependent aggregation kernels and initial conditions. The two techniques lead to identical predictions, with the former requiring only a fraction of the computational effort. The rule-set automatically reduces aggregation of particles of same composition to a 1-d problem. A midway change in the direction of expansion of domain, effected by the addition of particles of different mean composition, is captured correctly by the rule-set. The evolving shape of a computational domain carries with it the signature of the aggregation process, which can be insightful in complex and time dependent aggregation conditions. (c) 2012 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

An aggregation rule maps each profile of individual strict preference orderings over a set of alternatives into a social ordering over that set. We call such a rule strategyproof if misreporting one’s preference never produces a social ordering that is strictly between the original ordering and one’s own preference. After describing a few examples of manipulable rules, we study in some detail three classes of strategy-proof rules: (i)rules based on a monotonic alteration of the majority relation generated by the preference profile; (ii)rules improving upon a fixed status-quo; and (iii) rules generalizing the Condorcet-Kemeny aggregation method.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

For most of the work done in developing association rule mining, the primary focus has been on the efficiency of the approach and to a lesser extent the quality of the derived rules has been emphasized. Often for a dataset, a huge number of rules can be derived, but many of them can be redundant to other rules and thus are useless in practice. The extremely large number of rules makes it difficult for the end users to comprehend and therefore effectively use the discovered rules and thus significantly reduces the effectiveness of rule mining algorithms. If the extracted knowledge can’t be effectively used in solving real world problems, the effort of extracting the knowledge is worth little. This is a serious problem but not yet solved satisfactorily. In this paper, we propose a concise representation called Reliable Approximate basis for representing non-redundant approximate association rules. We prove that the redundancy elimination based on the proposed basis does not reduce the belief to the extracted rules. We also prove that all approximate association rules can be deduced from the Reliable Approximate basis. Therefore the basis is a lossless representation of approximate association rules.